Combining Performance Actions with Spectral Models for Violin Sound Transformation
نویسندگان
چکیده
In this work we present a violin timbre model that takes into account performance gestures. It is built by analysis of performance data using machine learning methods and it is able to predict the timbre given a set of performance actions. Gestural data and sound are synchronously captured by means of 3D motion trackers attached to the instrument and a bridge pickup. The model is used for sample transformation within a spectral concatenative synthesizer informed by gestures. INTRODUCTION Spectral concatenative synthesis models [1], [2] generate sound by concatenation of spectrally transformed samples. Sample concatenation is crucial for the quality of sound produced, and sometimes transitions between two samples do not sound natural, especially in the case of sustained excitation instruments such as the violin, because they have a wider timbre space and need a continuous control. One manner in which to improve these models in order to provide better controllability and expressive capabilities is to take into consideration performance gestures, that is, informing the model with "how is the instrument played". Performance actions are sound producing gestures articulated by the musician that control/drive the production of sound (see in [8] for a categorization of musical gestures). When performing with a violin, one can produce a wide range of different timbre variations, by applying a complex combination of actions controlled by bow and fingers. Bowing actions are the most relevant concerning timbre and therefore we will focus on them. We have developed a sensing system [14] by making use of two Polhemus 3D-motion trackers. Using data provided by this system we obtain bowing performance actions with great accuracy. Sound is acquired by means of a 4-channel bridge pickup that is then spectrally analyzed. With this setup, we are able to synchronously collect large amounts of performance data (gestures and sound), that is used to train a set of neural networks. The trained networks are finally used in the transformation stage of a spectral concatenative synthesizer. The paper is structured as follows: First we describe which data is acquired and how. In the case of sound recording, we discuss why are we using a bridge pickup instead of another device. Then we present the neural network that models the timbre, detailing its structure, inputs, output, the dataset for training and its performance. Finally we outline its use in the transformation procedure by the synthesizer and conclude by commenting some evaluation results and presenting further developments of the model. DATA AQUISITON With our measuring system consisting of the bridge pickup and the motion trackers we can capture an enormous amount of performance data. The main advantages over other systems like bowing machines [4] are that the range of the bowing actions is not constrained by the machine and that we can capture real performance data.
منابع مشابه
Gesture based synthesis of bowed string instruments
Synthesis of traditional music instruments has been an active research area and there exist successful implementations of instruments with low degree of control like non-sustained excitation instruments. But for instruments with sustained excitation, such as bowed strings or wind instruments, where the interaction between instrument and performer is continuous, the quality of existing models is...
متن کاملGesture Sampling For Instrumental Sound Synthesis: Violin Bowing As A Case Study
This paper presents a framework in which samples of bowing gesture parameters are retrieved and concatenated from a database of violin performances by attending to an annotated input score. Resulting bowing parameter signals are then used to synthesize sound by means of both a digital waveguide violin physical model, and an spectral-domain additive synthesizer.
متن کاملSpectral Modeling Synthesis: A Sound Analysis/Synthesis System Based on a Deterministic plus Stochastic Decomposition
When generating musical sound on a digital computer, it is important to have a good model whose parameters provide a rich source of meaningful sound transformations. Three basic model types are used widely today for musical sound generation: instrument models, spectrum models, and abstract models. Instrument models attempt to parameterize a sound at its source, such as a violin, clarinet, or vo...
متن کاملScore Level Timbre Transformations of Violin Sounds
The ability of a sound synthesizer to provide realistic sounds depends to a great extent on the availability of expressive controls. One of the most important expressive features a user of the synthesizer would desire to have control of, is timbre. Timbre is a complex concept related to many musical indications in a score such as dynamics, accents, hand position, string played, or even indicati...
متن کاملComparison of Subspace Analysis-based and Statistical Model-based Algorithms for Musical Instrument Classification
In this paper, three classes of algorithms for automatic classification of individual musical instrument sounds are compared. The first class of classifiers is based on Non-negative Matrix Factorization, the second class of classifiers employs automatic feature selection and Gaussian Mixture Models and the third is based on continuous Hidden Markov Models. Several perceptual features used in ge...
متن کاملذخیره در منابع من
با ذخیره ی این منبع در منابع من، دسترسی به آن را برای استفاده های بعدی آسان تر کنید
عنوان ژورنال:
دوره شماره
صفحات -
تاریخ انتشار 2007